# Reinforcement learning alignment

Llama 3 3 Nemotron Super 49B V1 FP8
Other
Llama-3.3-Nemotron-Super-49B-v1-FP8 is a large language model derived from Meta Llama-3.3-70B-Instruct, optimized to enhance reasoning capabilities, conversational preferences, and task execution, supporting a context length of 128K tokens.
Large Language Model Transformers English
L
nvidia
81
4
Llama 3.3 70B Instruct
Meta Llama 3.3 is a 70-billion-parameter multilingual large language model optimized for multilingual dialogue scenarios, supporting 8 languages and excelling in multiple benchmark tests.
Large Language Model Transformers Supports Multiple Languages
L
meta-llama
1.1M
2,282
Meta Llama 3 70B Instruct GGUF
GGUF quantized version based on Meta's official Llama 3 70B instruction fine-tuned model, supporting 2-16bit multiple quantization levels, suitable for locally deployed dialogue scenarios
Large Language Model English
M
MaziyarPanahi
18.89k
168
Xwin LM 7B V0.2
Xwin-LM is a large language model developed based on Llama2, focusing on alignment technology research and demonstrating excellent performance in the AlpacaEval benchmark.
Large Language Model Transformers
X
Xwin-LM
4,345
46
Llama 2 7b Chat Hf
Llama 2 is a series of generative text models open-sourced by Meta, including pre-trained and dialogue fine-tuned versions ranging from 7 billion to 70 billion parameters. This model is the 7 billion parameter dialogue-optimized version, fine-tuned with reinforcement learning for dialogue scenarios.
Large Language Model Transformers English
L
meta-llama
1.2M
4,390
Llama 2 13b Chat Hf
Llama 2 is Meta's open-source 13-billion-parameter conversation-optimized large language model, aligned with human preferences using RLHF, supporting 4k context length
Large Language Model Transformers English
L
meta-llama
147.25k
1,075
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase